AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
RAG Enhancement

# RAG Enhancement

Japanese Reranker Tiny V2
MIT
This is a very compact and fast Japanese reranking model, suitable for improving the accuracy of RAG systems and can run efficiently on CPUs or edge devices.
Text Embedding Japanese
J
hotchpotch
339
3
Qwen2.5 Aloe Beta 7B
Apache-2.0
Qwen2.5-Aloe-Beta-7B is an open-source large medical language model that achieves state-of-the-art performance in multiple medical tasks. It is fine-tuned based on the Qwen2.5-7B architecture, and the training data covers 1.8 billion tokens of diverse medical tasks.
Large Language Model Transformers English
Q
HPAI-BSC
631
5
Nemotron Mini 4B Instruct
Other
Nemotron-Mini-4B-Instruct is a response generation model developed by NVIDIA, optimized for role-playing, retrieval-augmented generation, and function calling. It is fine-tuned based on Minitron-4B-Base and supports a context length of 4096 tokens.
Large Language Model Transformers English
N
nvidia
674
162
Tinyllama Cinder Agent V1
MIT
A lightweight agent model fine-tuned based on TinyLlama, focusing on tool calls, RAG, and general assistant functions
Large Language Model Transformers
T
Josephgflowers
63
2
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase